GRIT-VLP: Grouped Mini-batch Sampling for Efficient Vision and Language Pre-training
نویسندگان
چکیده
Most of the currently existing vision and language pre-training (VLP) methods have mainly focused on how to extract align text features. In contrast mainstream VLP methods, we highlight that two routinely applied steps during crucial impact performance pre-trained model: in-batch hard negative sampling for image-text matching (ITM) assigning large masking probability masked modeling (MLM). After empirically showing unexpected effectiveness above steps, systematically devise our GRIT-VLP, which adaptively samples mini-batches more effective mining ITM while maintaining computational cost pre-training. Our method consists three components: 1) GRouped mIni-baTch (GRIT) strategy collects similar examples in a mini-batch, 2) ITC consistency loss improving ability, 3) enlarged MLM. Consequently, show GRIT-VLP achieves new state-of-the-art various downstream tasks with much less cost. Furthermore, demonstrate model is essentially par ALBEF, previous state-of-the-art, only one-third training epochs same data. Code available at https://github.com/jaeseokbyun/GRIT-VLP .
منابع مشابه
Balanced Mini-batch Sampling for SGD Using Determinantal Point Processes
We study a mini-batch diversification scheme for stochastic gradient descent (SGD). While classical SGD relies on uniformly sampling data points to form a mini-batch, we propose a non-uniform sampling scheme based on the Determinantal Point Process (DPP). The DPP relies on a similarity measure between data points and gives low probabilities to mini-batches which contain redundant data, and high...
متن کاملVariable Mini-Batch Sizing and Pre-Trained Embeddings
This paper describes our submission to the WMT 2017 Neural MT Training Task. We modified the provided NMT system in order to allow for interrupting and continuing the training of models. This allowed mid-training batch size decrementation and incrementation at variable rates. In addition to the models with variable batch size, we tried different setups with pre-trained word2vec embeddings. Asid...
متن کاملSubmodular Mini-Batch Training in Generative Moment Matching Networks
Generative moment matching network (GMMN), which is based on the maximum mean discrepancy (MMD) measure, is a generative model for unsupervised learning, where the mini-batch stochastic gradient descent is applied for the update of parameters. In this work, instead of obtaining a mini-batch randomly, each mini-batch in the iterations is selected in a submodular way such that the most informativ...
متن کاملBatch-Expansion Training: An Efficient Optimization Paradigm for Machine Learning
We propose Batch-Expansion Training (BET), a framework for running a batch optimizer on a gradually expanding dataset. As opposed to stochastic approaches, batches do not need to be resampled i.i.d. at every iteration, thus making BET more resource efficient in a distributed setting, and when disk-access is constrained. Moreover, BET can be easily paired with most batch optimizers, does not req...
متن کاملDistributed Mini-Batch SDCA
We present an improved analysis of mini-batched stochastic dual coordinate ascent for regularized empirical loss minimization (i.e. SVM and SVM-type objectives). Our analysis allows for flexible sampling schemes, including where data is distribute across machines, and combines a dependence on the smoothness of the loss and/or the data spread (measured through the spectral norm).
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Lecture Notes in Computer Science
سال: 2022
ISSN: ['1611-3349', '0302-9743']
DOI: https://doi.org/10.1007/978-3-031-19800-7_23